Weeks 07 & 08
April 20, 2026
Households maximize utility \((u)\) over time
Utility depends on consumption \((c)\) and on hours worked \((\ell)\): \[u(c_t , \ell_t)=\frac{c_t^{1-\sigma}}{1-\sigma}- \theta \ell_t \tag{1}\]
Intertemporal utility is discounted by a factor \(\beta\), then \[ u(\cdot)=\overbrace{\beta^0 \cdot u\left(c_{0}, \ell_{0}\right)}^{\text {period } 0}+\overbrace{\beta^1 \cdot u\left(c_{1}, \ell_{1}\right)}^{\text {period } 1}+\overbrace{\beta^2 \cdot u\left(c_{2}, \ell_{2}\right)}^{\text {period } 2}+\ldots+\overbrace{\beta^T \cdot u\left(c_{T}, \ell_{T}\right)}^{\text {period } T} \]
The discount factor is \(\beta=1 /(1+r)\); while \(r\) is the subjective discount rate of future utility.
Each firm produces goods & services with the following production function: \[ y_t=a_t k_{t}^\alpha \ell_t^{1-\alpha}\tag{2} \]
Firms maximize profits and take wages and rental prices as given: \[max \{\pi = y_t - w_t \ell_t - r_t k_t \} \tag{3} \] where \(w_t\) is the wage rate, \(r_t\) is the rental price of capital.
As markets are fully competitive, factor returns are equal to their marginal products: \[w_t = \frac{\partial y_t}{\partial \ell_t} \quad \text{and} \quad r_t = \frac{\partial y_t}{\partial k_t} \tag{4}\]
Population remains constant over time. This means that the total number of households remains constant: \[n_t = \overline{n} \tag{5}\]
Therefore, as far as labor is concerned, only changes in the average hours worked per household (\(\ell\)) can affect the level of output.
Capital accumulation is given by the usual definition from national accounts: \[ k_{t+1} \equiv i_t+(1-\delta) k_t \tag{6} \]
\(k\) represents capital per household, \(i\) investment per household, and \(\delta\) the depreciation rate.
Assume that technology \((a_t)\) does not increase over time (there is no trend)
\(a_t\) fluctuates around its steady state value \(\left(\overline{a}\right)\), due to exogenous shocks \(\left(\varepsilon_t\right)\) \[ \ln a_t=(1-\rho) \ln \overline{a}+\rho \ln a_{t-1}+\varepsilon_t \quad, \quad \rho<1 \tag{7} \]
Why logarithms? To make things easier!
Define: \[\hat{a}_t=\ln a_t-\ln \overline{a}\]
Then eq. (7) can be written as \[ \hat{a}_t=\rho \cdot \hat{a}_{t-1}+\varepsilon_t \] i.e., the log-deviation of technology from its steady state is an AR(1) process with \(\rho<1\), and mean=zero.
In a simple economy, output is either consumed by households (\(c_t\)) or invested in new capital (\(i_t\)): \[y_t \equiv c_t + i_t \tag{8}\]
We know that output is produced according to eq. (2): \[y_t=a_t k_{t}^\alpha \ell_t^{1-\alpha}\tag{2'}\]
And capital accumulation is given by eq. (6): \[k_{t+1} \equiv i_t+(1-\delta) k_t \tag{6'}\]
Inserting eq. (2’) & (6’) into (8) we get the overall macroeconomic constraint as: \[a_t k_{t}^\alpha \ell_t^{1-\alpha}= c_t + k_{t+1} - (1-\delta) k_t \tag{8'}\]
The intertemporal maximization of utility \[ \mathcal{L} = \sum_{t=0}^\infty \beta^t \Big\{ \ \underbrace{u(c_t,\ell_t)}_{\text{utility}} + \lambda_t \ \underbrace{ (a_t k_t^\alpha \ell_t^{1-\alpha} + (1-\delta)k_t - c_t - k_{t+1}) }_{\text{resource constraint}} \Big\} \tag{10} \] where \(\lambda_t\) stands for the Lagrangian multiplier
First Order Conditions (FOCs):
It looks complicated, but it is not! What you need is PATIENCE!
\[ {\color{teal}\partial \mathcal{L} / \partial c_t=\beta^0\left(u_{c_t}^{\prime}-\lambda_t\right)=0} \qquad \qquad \]
\[ {\color{blue}\partial \mathcal{L} / \partial k_{t+1}=-\beta^0 \cdot \lambda_t+\beta^1 \cdot \lambda_{t+1}(\underbrace{\alpha \cdot a_{t+1} k_{t+1}^{\alpha-1} \ell_{t+1}^{1-\alpha} +1-\delta}_{\equiv \ r_{t+1}})=0} \]
\[ {\color{red}\partial \mathcal{L} / \partial \ell_t=\beta^0[u_{\ell_t}^{\prime}+\lambda_t(1-\alpha) \underbrace{a_t k_t^\alpha \ell_t^{-\alpha}}_{=\ y_t / \ell_t}]=0} \]
\[ {\color{black}\partial \mathcal{L} / \partial \lambda_t=\beta^0\left(a_t k_t^\alpha \ell_t^{1-\alpha}+(1-\delta) k_t-c_t-k_{t+1}\right)=0} \]
\[ \begin{aligned} {\color{teal}\partial\mathcal{L}/\partial c_t} &= {\color{teal}\beta^0(u'_{c_t}-\lambda_t)=0} && \text{(FOC1)} \\[1.2em] {\color{blue}\partial\mathcal{L}/\partial k_{t+1}} &= {\color{blue}-\beta^0\lambda_t + \beta^1\lambda_{t+1} \Bigl(\underbrace{\alpha \cdot a_{t+1}k_{t+1}^{\alpha-1}\ell_{t+1}^{1-\alpha}+1 - \delta}_{\equiv \ r_{t+1}}\Bigr)=0} && \text{(FOC2)} \\[0.1em] {\color{red}\partial\mathcal{L}/\partial \ell_t} &= {\color{red}\beta^0\Bigl[u'_{\ell_t} +\lambda_t(1-\alpha)\underbrace{a_t k_t^\alpha\ell_t^{-\alpha}}_{= \ y_t/\ell_t}\Bigr]=0} && \text{(FOC3)} \\[0.1em] {\color{black}\partial\mathcal{L}/\partial\lambda_t} &= \beta^0\Bigl(a_t k_t^\alpha \ell_t^{1-\alpha} + (1-\delta)k_t - c_t - k_{t+1}\Bigr)=0 && \text{(FOC4)} \end{aligned} \]
From FOC1, we know that: \[\beta^0(u'_{c_t}-\lambda_t)=0 \Rightarrow \lambda_t = u'_{c_t} \tag{11}\]
As from eq. (1), the marginal utility of consumption can be written as: \[u'(c_t) = \frac{\partial u(c_t, \ell_t)}{\partial c_t} = c_t^{-\sigma} \tag{12} \]
By inserting eq. (12) into (11), we get: \[ \lambda_t = c_t^{-\sigma} \quad \tag{13} \]
Obviously, if \(\lambda_t = c_t^{-\sigma}\), then \[\lambda_{t+1} = c_{t+1}^{-\sigma} \tag{14}\]
From FOC2, we know that: \(\quad \color{blue}{\beta^0\lambda_t = \beta^1\lambda_{t+1} r_{t+1}}\).
By inserting eq. (13) and (14) into this FOC2, we get: \[c_t^{-\sigma}=\beta \cdot c_{t+1}^{-\sigma} \cdot r_{t+1} \tag{15}\]
This is the famous Euler equation. It gives the optimal level of consumption over time, which depends on the discount factor \(\beta\), the risk aversion parameter \(\sigma\) and the net return on capital \(r_{t+1}\).
Notice that the Euler equation can be written as a ratio \((\xi)\): \[\xi \equiv \frac{c_{t+1}}{c_t} = \left(\beta \cdot r_{t+1}\right)^{1/\sigma} \tag{16}\] Implying that: \(\partial \xi /\partial r_{t+1} < 0\), \(\partial \xi /\partial \beta < 0\), \(\partial \xi /\partial \sigma < 0\).
From FOC3, we know that: \(\qquad {\color{red}\beta^0\Bigl[u'_{\ell_t} +\lambda_t(1-\alpha)\underbrace{a_t k_t^\alpha\ell_t^{-\alpha}}_{= \ y_t/\ell_t}\Bigr]=0}\)
As we know that from eq. (1) the marginal utility of working is given by: \[u'(\ell_t) = \frac{\partial u(c_t, \ell_t)}{\partial \ell_t} = -\theta \tag{17}\]
By inserting eq. (13) and (17) into (FOC3), we get: \[-\theta + c_t^{-\sigma} (1-\alpha) \frac{y_t}{\ell_t} = 0 \quad \Rightarrow \quad \frac{y_t}{\ell_t}=\left(\frac{\theta}{1-\alpha}\right) c_t^\sigma \tag{18}\]
This is the static equation that optimizes the labour supply.
From FOC4, we know that: \[{\color{black}\beta^0\Bigl(a_t k_t^\alpha \ell_t^{1-\alpha} + (1-\delta)k_t - c_t - k_{t+1}\Bigr)=0} \tag{20}\]
As \(\beta^0 =1\), this implies that the resource constraint is fully satisfied: \[a_t k_t^\alpha \ell_t^{1-\alpha} = c_t + k_{t+1} - (1-\delta)k_t \tag{21}\]
No more simplifications are possible in this equation.
\[ \begin{align} c_t^{-\sigma}&=\beta \cdot c_{t+1}^{-\sigma} \cdot r_{t+1} \tag{15'} \\[0.5em] \frac{y_t}{\ell_t}&=\left(\frac{\theta}{1-\alpha}\right) c_t^\sigma \tag{18'} \\[0.5em] \underbrace{a_t k_t^\alpha \ell_t^{1-\alpha}}_{= \ y_t} &= c_t + \underbrace{k_{t+1} - (1-\delta)k_t}_{= \ i_t} \tag{21'} \end{align} \]
\[ \begin{align} r_{t+1} &= \alpha a_{t+1} k_{t+1}^{\alpha-1} \ell_{t+1}^{1-\alpha} + 1 - \delta \tag{ROC} \\[0.5em] y_t &= a_t k_t^\alpha \ell_t^{1-\alpha}\tag{2'} \\[0.5em] \ln a_t &= (1-\rho)\ln\overline{a} + \rho\ln a_{t-1} + \varepsilon_t \tag{7'} \\[0.5em] y_t &= c_t + i_t\tag{8'} \end{align} \]
The solution to the model involves 7 non-linear equations and 7 variables \(\{c_t, \ell_t, k_t, y_t, a_t, r_{t+1}, i_t\}\):
\[ \begin{align} c_t^{-\sigma}&=\beta \cdot c_{t+1}^{-\sigma} \cdot r_{t+1} \tag{15'} \\[0.5em] \frac{y_t}{\ell_t}&=\left(\frac{\theta}{1-\alpha}\right) c_t^\sigma \tag{18'} \\[0.5em] r_{t+1} & \equiv \alpha \frac{y_{t+1}}{k_{t+1}} + 1 - \delta \tag{ROC} \\[0.5em] y_t &= a_t k_t^\alpha \ell_t^{1-\alpha}\tag{2'} \\[0.5em] k_{t+1} & \equiv (1-\delta)k_t + i_t \tag{6'}\\[0.5em] \ln a_t &= (1-\rho)\ln\overline{a} + \rho\ln a_{t-1} + \varepsilon_t \tag{7'} \\[0.5em] y_t & \equiv c_t + i_t\tag{8'} \end{align} \]
To compute the steady-state of a variable \(x\), impose the usual condition: \[x_{t+1} = x_t = \overline{x}\]
Let us start with the Euler equation (eq. 15): \[ \begin{align*} c_t^{-\sigma} & =\beta \left(c_{t+1}^{-\sigma} \cdot r_{t+1}\right) \Rightarrow \frac{(\overline{c})^{-\sigma}}{(\overline{c})^{-\sigma}} =\beta \cdot \overline{r} \\ \overline{r} & =\beta^{-1} \tag{22} \end{align*} \]
Using the (ROC) definition \(r_{t+1} \equiv \alpha \frac{y_{t+1}}{k_{t+1}}+1-\delta\), and the result in eq. (22), we get: \[ \beta^{-1} \equiv \alpha\left(\frac{\bar{y}}{\bar{k}}\right)+1-\delta \quad \Rightarrow \frac{\bar{y}}{\bar{k}} =\frac{\beta^{-1}+\delta-1}{\alpha} \tag{23} \]
From eq. (6’), we can obtain: \[ \overline{k} =(1-\delta) \overline{k}+\bar{i} \quad \Rightarrow \quad \frac{\overline{i}}{\overline{k}} =\delta \tag{24}\]
From eq. (23) we know that \(\frac{\bar{y}}{\bar{k}}= \frac{\beta^{-1}+\delta-1}{\alpha}\), and from (24) we have \(\frac{\overline{i}}{\overline{k}} =\delta\). Therefore, we can obtain after some manipulations: \[ \frac{\overline{i}}{\overline{y}}=\frac{\frac{\bar{i}}{\bar{k}}}{\frac{\bar{y}}{\bar{k}}}=\phi \quad , \quad \text {with } \quad \phi\equiv \frac{\alpha \delta}{\beta^{-1}+\delta-1} \tag{25} \]
From eq.(8’) we have \(y_t=c_t+i_t\). By dividing both sides by \(y_t\), and knowing that from (25) we have \(\frac{\overline{i}}{\overline{y}}=\phi\), we get: \[ \frac{\overline{c}}{\overline{y}}=1-\frac{\overline{i}}{\overline{y}}=1-\phi \tag{26} \]
From eq. (18’), we can obtain: \[ \frac{\overline{y}}{\overline{\ell}} =\frac{\theta}{1-\alpha} \left (\overline{c}\right)^\sigma \tag{27} \]
Finally, by construction, without the shocks, we can obtain the following in the steady-state: \[ \ln \overline{a} =(1-\rho)\ln\overline{a} + \rho\ln \overline{a} \quad \Rightarrow \quad 0=0 \tag{28} \]
Suppose we have a nonlinear function \[y=f(x) = \frac{1}{4}(4+x-x^2)\]
We want to approximate it around a point \(\bar{x}=1\).
We can approximate it by its tangent line at \(\bar{x}=1\), given by: \[y = f(\bar{x}) + f'(\bar{x})(x-\bar{x})=-1/4(x + 5/4)\]
See next figure
\(~~~~\)The original non-linear model \(~~~~~~~~~~~~~~~~~~~~~~~~\) The linearized model
\[ \begin{array}{ccc} c_t^{-\sigma}=\beta \left(c_{t+1}^{-\sigma} r_{t+1}\right) & \Leftrightarrow & \hat{c}_t= \hat{c}_{t+1}-\frac{1}{\sigma} \hat{r}_{t+1} && \text{ (L1)} \\ y_t / \ell_t=[\theta /(1-\alpha)] c_t^\sigma & \Leftrightarrow & \hat{\ell}_t=\hat{y}_t-\sigma \hat{c}_t && \text{ (L2)} \\ k_{t+1}=(1-\delta) k_{t}+i_t & \Leftrightarrow & \hat{k}_{t+1}=(1-\delta) \hat{k}_{t}+\hat{i}_t \color{red}{\frac{\overline {i}}{\overline{k}}} && \text{ (L3)} \\ y_t=a_t k_{t}^\alpha \ell_t^{1-\alpha} & \Leftrightarrow & \hat{y}_t=\hat{a}_t+\alpha \hat{k}_{t}+(1-\alpha) \hat{\ell}_t && \text{ (L4)} \\ c_t+i_t=y_t & \Leftrightarrow & \hat{y}_t=\hat{c}_t {\color{red}{\frac{\overline{c}}{\overline{y}}}} +\hat{i}_t {\color{red}{\frac{\overline{i}}{\overline{y}}}} && \text{ (L5)} \\ r_{t+1} \equiv \alpha\left(y_{t+1} / k_{t+1}\right)+1-\delta & \Leftrightarrow & \hat{r}_{t+1}=\frac{\alpha {\color{red}{\left(\overline{y} / \overline{k}\right)}}}{\alpha {\color{red}{\left(\overline{y} / \overline{k}\right)}} +1-\delta} \color{blue}{\hat{z}_{t+1}} && \text{ (L6)} \\ \mathrm{ln} a_t=(1-\rho) \ln \overline{a}+\rho \ln a_{t-1}+\varepsilon_t & \Leftrightarrow & \hat{a}_t=\rho \hat{a}_{t-1}+\varepsilon_t && \text{ (L7)} \\ \end{array} \]
\[ \begin{array}{ccc} \hat{c}_t= \hat{c}_{t+1}-\frac{1}{\sigma} \hat{r}_{t+1} & \Leftrightarrow & \hat{c}_t= \mathbb{E}_t\hat{c}_{t+1}-\frac{1}{\sigma} \mathbb{E}_t \hat{r}_{t+1} && \text{ (S1)}\\ \hat{\ell}_t=\hat{y}_t-\sigma \hat{c}_t & \Leftrightarrow & \hat{\ell}_t=\hat{y}_t-\sigma \hat{c}_t && \text{ (S2)}\\ \hat{k}_{t+1}=(1-\delta) \hat{k}_{t}+\hat{i}_t \color{red}{\frac{\overline {i}}{\overline{k}}} & \Leftrightarrow & \hat{k}_{t+1}=(1-\delta) \hat{k}_{t}+ \hat{i}_t \cdot \color{red}{\delta} && \text{ (S3)}\\ \hat{y}_t=\hat{a}_t+\alpha \hat{k}_{t}+(1-\alpha) \hat{\ell}_t & \Leftrightarrow & \hat{y}_t=\hat{a}_t+\alpha \hat{k}_{t}+(1-\alpha) \hat{\ell}_t && \text{ (S4)}\\ \hat{y}_t=\hat{c}_t {\color{red}{\frac{\overline{c}}{\overline{y}}}}+\hat{i}_t \color{red}{\frac{\overline{i}}{\overline{y}}} & \Leftrightarrow & \hat{y}_t=\hat{c}_t {\color{red}{\left(1-\phi\right)}}+\hat{i}_t \cdot \color{red}{\phi} && \text{ (S5)}\\ \hat{r}_{t+1}=\frac{\alpha \color{red}{\left(\overline{y} / \overline{k}\right)}}{\alpha {\color{red}{\left(\overline{y} / \overline{k}\right)}} +1-\delta} \color {blue}{\hat{z}_{t+1}} & \Leftrightarrow & \mathbb{E}_t\hat{r}_{t+1}= \color{red}{\left(1+\beta \delta- \beta \right)} \color {blue}{\hat{z}_{t+1}} && \text{ (S6)}\\ \hat{a}_t=\rho \hat{a}_{t-1}+\varepsilon_t & \Leftrightarrow & \hat{a}_t=\rho \hat{a}_{t-1}+\varepsilon_t && \text{ (S7)}\\ \end{array} \]
In (S6) we define \(\small{\color{blue}{\hat{z}_{t+1} \equiv \hat{a}_{t+1}+(\alpha-1) \hat{k}_{t+1} + (1-\alpha) \mathbb{E}_t \hat{\ell}_{t+1}}}\)
By inserting (S6) into (S1) we can obtain: \[ \hat{c}_t=\mathbb{E}_t \hat{c}_{t+1}-\frac{\varphi}{\sigma} \left[\hat{a}_{t+1}+(\alpha-1) \hat{k}_{t+1}+(1-\alpha) \mathbb{E}_t \hat{\ell}_{t+1}\right] \ , \quad \varphi \equiv 1+\beta \delta-\beta \]
Equalizing (S4) & (S5), solving for \(\hat{i}_t\), and inserting this result into (S3), we get: \[ \hat{k}_{t+1}= \frac{1}{\beta}\hat{k}_t+\frac{\delta}{\phi} \hat{a}_t-\frac{\delta(1-\phi)}{\phi} \hat{c}_t + \frac{\delta (1-\alpha)}{\phi} \hat{\ell}_t \ \ , \quad \phi \equiv \frac{\alpha \delta}{\beta^{-1}+\delta-1} \]
We have 2 equations x 4 variables: \(\hat{c}_t, \hat{k}_t, \hat{a}_t, \hat{\ell}_t\). We need to bring in two more equations. Insert (S4) into (S2) and we get: \[ \mathbb{E}_t \hat{\ell}_{t+1}= \frac{1}{\alpha} \left( \hat{a}_{t+1}+\alpha \hat{k}_{t+1} -\sigma \mathbb{E}_t \hat{c}_{t+1} \right) \\ \]
The last one is the law of motion for the technology: \[\hat{a}_{t+1}=\rho \hat{a}_{t}+\varepsilon_{t+1}\]
| Variables | Equations |
|---|---|
| Technology | \(\hat{a}_{t+1} = \rho \cdot \hat{a}_t + \varepsilon_{t+1}\) |
| Capital | \(\hat{k}_{t+1}= \frac{1}{\beta}\hat{k}_t+\frac{\delta}{\phi} \hat{a}_t-\frac{\delta(1-\phi)}{\phi} \hat{c}_t + \frac{\delta (1-\alpha)}{\phi} \hat{\ell}_t \ \ , \ \ \phi \equiv \frac{\alpha \delta}{\beta^{-1}+\delta-1}\) |
| Labor | \(\mathbb{E}_t \hat{\ell}_{t+1}=\frac{1}{\alpha}\left(\hat{a}_{t+1}+\alpha \hat{k}_{t+1} -\sigma \mathbb{E}_t \hat{c}_{t+1}\right)\) |
| Consumption | \(\hat{c}_t=\mathbb{E}_t \hat{c}_{t+1}-\frac{\varphi}{\sigma}\left[\hat{a}_{t+1}+(\alpha-1) \hat{k}_{t+1}+(1-\alpha) \mathbb{E}_t \hat{\ell}_{t+1}\right] \ \ , \ \ \varphi \equiv 1+\beta \delta-\beta\) |
The model can be written in state space form as: \[
\mathcal{A}\left[\begin{array}{l}w_{t+1} \\
\mathbb{E}_{t} v_{t+1}
\end{array}\right]=\mathcal{B}\left[\begin{array}{l}w_{t} \\
v_{t}
\end{array}\right]+\mathcal{C}\left[\begin{array}{c}
\varepsilon_{t+1}^{w} \\
\varepsilon_{t+1}^{v}
\end{array}\right]+\mathcal{D} \tag{BK1}
\]
\(w\) is a vector of backward-looking variables (it may include static vriables as well), and \(v\) is a vector of forward looking variables.
Multiplying both sides of (BK1) by \(\mathcal{A}^{-1}\), leads to:
\[\left[\begin{array}{c}
w_{t+1} \\
\mathbb{E}_{t} v_{t+1}
\end{array}\right]=\underbrace{\mathcal{A}^{-1} \mathcal{B}}_{\mathcal{R}}\left[\begin{array}{c}
w_{t} \\
v_{t+1}
\end{array}\right]+\underbrace{\mathcal{A}^{-1} \mathcal{C}}_{\mathcal{U}}\left[\begin{array}{c}
\varepsilon_{t+1}^{w} \\
\varepsilon_{t+1}^{v}
\end{array}\right] + \underbrace{\mathcal{A}^{-1} \mathcal{D}}_{\mathcal{H}} \tag{BK2}
\]
The BK conditions: for a model to have a unique and stable solution, the matrix \(\mathcal{R}\) has \((m,n)\) eigenvalues such that \(|m_v|>1\) and \(|n_w|<1\).
The Jordan decomposition was the algebra technique used by Blanchard-Kahn (1980) to solve a DSGE model.
Suppose we have a square matrix \(\mathcal{R}\)
The Jordan decomposition of \(\mathcal{R}\) is given by: \[ \mathcal{R}=P \Lambda P^{-1} \]
\(P\) contains as columns the eigenvectors of \(\mathcal{R}\)
\(\Lambda\) is a diagonal matrix containing the eigenvalues of \(\mathcal{R}\) in the main diagonal.
\(P^{-1}\) is the inverse of \(P\)
Blanchard, O. J., & Kahn, C. M. (1980). The Solution of Linear Difference Models under Rational Expectations. Econometrica, 48(5), 1305–1311.
Our system was given by: \[ \qquad \qquad \qquad \left[\begin{array}{c} w_{t+1} \\ \mathbb{E}_t v_{t+1} \end{array}\right]=\mathcal{R}\left[\begin{array}{l} w_t \\ v_t \end{array}\right]+\mathcal{U}\left[\begin{array}{c} \varepsilon_{t+1}^w \\ \varepsilon_{t+1}^v \end{array}\right] \qquad \qquad \qquad \qquad \tag{BK3} \]
Apply the Jordan decomposition \(\mathcal{R}=P \Lambda P^{-1}\) to (BK3), and we get: \[ \qquad \qquad \left[\begin{array}{c} w_{t+1} \\ \mathbb{E}_t v_{t+1} \end{array}\right]=P \Lambda P^{-1}\left[\begin{array}{l} w_t \\ v_t \end{array}\right]+\mathcal{U}\left[\begin{array}{l} \varepsilon_{t+1}^w \\ \varepsilon_{t+1}^v \end{array}\right] \qquad \qquad \qquad\tag{BK4} \]
Multiply both sides of (BK4) by \(P^{-1}\) : \[ P^{-1}\left[\begin{array}{c} w_{t+1} \\ \mathbb{E}_t v_{t+1} \end{array}\right]=\Lambda P^{-1}\left[\begin{array}{c} w_t \\ v_t \end{array}\right]+\underbrace{P^{-1} \mathcal{U}}_{\mathcal{M}} \cdot\left[\begin{array}{c} \varepsilon_{t+1}^w \\ \varepsilon_{t+1}^v \end{array}\right] \tag{BK5} \]
Next we have to apply a partition of the matrices above: \(P^{-1}, \Lambda, \mathcal{M}\).
Assume that there are no shocks affecting the forward-looking block: \[ \varepsilon_t^v=0, \forall t \]
Next, apply a partition to the matrices: \(P^{-1}, \Lambda, \mathcal{M}\): \[ \underbrace{\left[\begin{array}{ll} P_{11} & P_{12} \\ P_{21} & P_{22} \end{array}\right]\left[\begin{array}{c} w_{t+1} \\ \mathbb{E}_t v_{t+1} \end{array}\right]}_{\mathbb{E}_t\left[\begin{array}{c} \widetilde{v}_{t+1} \\ \bar{v}_{t+1} \end{array}\right]}=\left[\begin{array}{cc} \Lambda_1 & 0 \\ 0 & \Lambda_2 \end{array}\right] \underbrace{\left[\begin{array}{ll} P_{11} & P_{12} \\ P_{21} & P_{22} \end{array}\right]\left[\begin{array}{c} w_t \\ v_t \end{array}\right]}_{\left[\begin{array}{c} \widetilde{w}_t \\ \bar{v}_t \end{array}\right]}+\underbrace{\left[\begin{array}{ll} \mathcal{M}_{11} & \mathcal{M}_{12} \\ \mathcal{M}_{21} & \mathcal{M}_{22} \end{array}\right]}_M\left[\begin{array}{c} \varepsilon_{t+1}^w \\ 0 \end{array}\right] \]
Our transformed model looks much easier now: \[ \left[\begin{array}{c} \widetilde{w}_{t+1} \\ \mathbb{E}_t \tilde{v}_{t+1} \end{array}\right]=\left[\begin{array}{cc} \Lambda_1 & 0 \\ 0 & \Lambda_2 \end{array}\right]\left[\begin{array}{c} \widetilde{w}_t \\ \tilde{v}_t \end{array}\right]+M_{11} \cdot \varepsilon_{t+1}^w \qquad \qquad \qquad \tag{BK6} \]
Using these partitions, the solution is given in the next slide
To solve the model, we have to apply the following analytical solutions, first for the non-forward-looking block: \[w_{t+1}^{*}=\underbrace{\left[G^{-1} \Lambda_{1} G\right]}_{g} \cdot w_{t}^{*}+\underbrace{\left[G^{-1} M_{11}\right]}_{h} \cdot \varepsilon_{t+1} \tag{BK7}\]
and, then, for the forward-looking block: \[v_{t}^{*}=\underbrace{\left[-P_{22}^{-1} P_{21}\right]}_{f} \cdot w_{t}^{*} \tag{BK8}\]
with \[G \equiv P_{11}-P_{12}\left(P_{22}\right)^{-1} P_{21} \tag{BK9}\]
Therefore, the solution requires the computation of the matrices: \(g, h, f,G.\)
\[ \begin{aligned} &{\color{teal} \underbrace{ \begin{bmatrix} 1 & 0 & 0 & 0 \\ 0 & 1 & 0 & 0 \\ -\frac{1}{\alpha} & -1 & 1 & \frac{\sigma}{\alpha} \\ -\frac{\varphi}{\sigma} & -\frac{\varphi(\alpha-1)}{\sigma} & -\frac{\varphi(1-\alpha)}{\sigma} & 1 \end{bmatrix} }_{\mathcal A} \begin{bmatrix} \hat a_{t+1} \\ \hat k_{t+1} \\ \mathbb E_t \hat{\ell}_{t+1} \\ \mathbb E_t \hat{c}_{t+1} \end{bmatrix} }= \\[1em] &\underbrace{ \begin{bmatrix} \rho & 0 & 0 & 0 \\ \frac{\delta}{\phi} & \frac{1}{\beta} & \frac{\delta(1-\alpha)}{\phi} & -\frac{\delta(1-\phi)}{\phi} \\ 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 1 \end{bmatrix} }_{\mathcal B} \begin{bmatrix} \hat a_t \\ \hat k_t \\ \hat\ell_t \\ \hat c_t \end{bmatrix} + {\color{magenta} \underbrace{ \begin{bmatrix} 1 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 \end{bmatrix} }_{\mathcal C} \begin{bmatrix} \varepsilon^a_{t+1}\\ \varepsilon^k_{t+1}\\ \varepsilon^\ell_{t+1}\\ \varepsilon^c_{t+1} \end{bmatrix} } + \underbrace{ \begin{bmatrix} \color{red} 0\\ \color{red} 0\\ \color{red} 0\\ \color{red} 0 \end{bmatrix} }_{\mathcal D} \end{aligned} \]
This is an introduction to the Real Business Cycle model. Therefore, we do not expect students to cover all the fundamental aspects associated with a sophisticated model in modern macroeconomics.
For example, the linearization of the model can be left for further studies in more advanced courses. Everything else is not very complicated, and students should be able to understand it.
The study of this model can be understood from three different perspectives: (i) the basic techniques, (ii) technical details that go beyond the basic issues, and (iii) a critical appraisal of theoretical and empirical issues.
In this course, we concentrate only on the basic issues. Therefore, students should read the lecture notes by Mendes, V. (2026). The Real Business Cycle Model: An Introduction here. Students can skip the section about “linearization”.
For those who want to delve deeper into the model and the associated techniques, we do not know of anything better than an old set of notes by Krueger, D. (2007). Quantitative Macroeconomics: An Introduction, University of Pennsylvania, here. They are old but may still be the best for this task.
If students want to get a feeling of major theoretical and empirical controversies about this model (as well as more recent developments), an excellent source of information is the chapter by Mitman, K. (2025). “Chapter 14: Real Business Cycles”, in the online/huge textbook “Macroeconomics”, by Azzimonti, M. et al. (2025), available here.